2025-04-02 03:36:32,032 [ 133870 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-02 03:36:32,032 [ 133870 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-02 03:36:32,032 [ 133870 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-02 03:36:32,032 [ 133870 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_41er5q --privileged --dns-search='.' --memory=30709030912 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_access_control_on_cluster/test.py::test_access_control_on_cluster test_access_control_on_cluster/test.py::test_grant_all_on_cluster test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_alter_settings_on_cluster/test.py::test_default_database_on_cluster test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated 'test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0]' 'test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576]' test_backup_log/test.py::test_backup_log 'test_backup_restore_keeper_map/test.py::test_on_cluster[0]' 'test_backup_restore_keeper_map/test.py::test_on_cluster[1]' test_backup_restore_new/test_cancel_backup.py::test_cancel_backup test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup test_backup_restore_on_cluster/test_different_versions.py::test_different_versions test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node 'test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3]' 'test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket]' test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain]' test_backup_restore_s3/test.py::test_backup_to_s3 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single]' test_backup_restore_s3/test.py::test_backup_to_s3_multipart test_backup_restore_s3/test.py::test_backup_to_s3_named_collection 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3]' 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket]' 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable]' test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart test_backup_restore_s3/test.py::test_backup_to_tar test_backup_restore_s3/test.py::test_backup_to_tar_bz2 test_backup_restore_s3/test.py::test_backup_to_tar_gz test_backup_restore_s3/test.py::test_backup_to_tar_lzma test_backup_restore_s3/test.py::test_backup_to_tar_xz test_backup_restore_s3/test.py::test_backup_to_tar_zst test_backup_restore_s3/test.py::test_backup_to_zip 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True]' test_backup_restore_s3/test.py::test_incremental_backup_append_table_def test_backup_restore_s3/test.py::test_user_specific_auth test_catboost_evaluate/test.py::testAmazonModelManyRows test_catboost_evaluate/test.py::testAmazonModelSingleRow test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString test_catboost_evaluate/test.py::testConstantFeatures test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric test_catboost_evaluate/test.py::testInvalidLibraryPath test_catboost_evaluate/test.py::testInvalidModelPath test_catboost_evaluate/test.py::testModelPathIsNotAConstString test_catboost_evaluate/test.py::testModelUpdate test_catboost_evaluate/test.py::testNonConstantFeatures test_catboost_evaluate/test.py::testOnLowCardinalityFeatures test_catboost_evaluate/test.py::testOnNullableFeatures test_catboost_evaluate/test.py::testRecoveryAfterCrash test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop test_codec_encrypted/test.py::test_different_keys test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_reloader_interval/test.py::test_reload_config test_config_xml_full/test.py::test_xml_full_conf test_config_xml_main/test.py::test_xml_main_conf -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache Test order randomisation NOT enabled. Enable with --random-order or --random-order-bucket= rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: timeout-2.3.1, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0, random-order-1.1.1 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] test_composable_protocols/test.py::test_connections test_async_load_databases/test.py::test_async_load_system_database test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_catboost_evaluate/test.py::testAmazonModelManyRows test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup test_attach_partition_using_copy/test.py::test_all_replicated test_access_control_on_cluster/test.py::test_access_control_on_cluster test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] [gw8] PASSED test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 [gw8] PASSED test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 [gw8] PASSED test_composable_protocols/test.py::test_proxy_1 [gw3] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_backup_restore_new/test_cancel_backup.py::test_cancel_backup [gw3] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility [gw6] PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes [gw3] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node [gw7] PASSED test_access_control_on_cluster/test.py::test_access_control_on_cluster test_access_control_on_cluster/test.py::test_grant_all_on_cluster [gw7] PASSED test_access_control_on_cluster/test.py::test_grant_all_on_cluster test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster [gw7] PASSED test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore [gw6] PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node [gw3] PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile [gw0] PASSED test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] [gw3] PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_alter_settings_on_cluster/test.py::test_default_database_on_cluster [gw1] PASSED test_catboost_evaluate/test.py::testAmazonModelManyRows test_catboost_evaluate/test.py::testAmazonModelSingleRow test_config_decryption/test.py::test_successful_decryption_xml [gw1] PASSED test_catboost_evaluate/test.py::testAmazonModelSingleRow test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString [gw0] PASSED test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk [gw1] PASSED test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString test_catboost_evaluate/test.py::testConstantFeatures [gw1] PASSED test_catboost_evaluate/test.py::testConstantFeatures test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric [gw1] PASSED test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric test_catboost_evaluate/test.py::testInvalidLibraryPath [gw1] PASSED test_catboost_evaluate/test.py::testInvalidLibraryPath test_catboost_evaluate/test.py::testInvalidModelPath [gw1] PASSED test_catboost_evaluate/test.py::testInvalidModelPath test_catboost_evaluate/test.py::testModelPathIsNotAConstString [gw1] PASSED test_catboost_evaluate/test.py::testModelPathIsNotAConstString test_catboost_evaluate/test.py::testModelUpdate [gw1] PASSED test_catboost_evaluate/test.py::testModelUpdate test_catboost_evaluate/test.py::testNonConstantFeatures [gw6] PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node [gw1] PASSED test_catboost_evaluate/test.py::testNonConstantFeatures test_catboost_evaluate/test.py::testOnLowCardinalityFeatures [gw1] PASSED test_catboost_evaluate/test.py::testOnLowCardinalityFeatures test_catboost_evaluate/test.py::testOnNullableFeatures [gw1] PASSED test_catboost_evaluate/test.py::testOnNullableFeatures test_catboost_evaluate/test.py::testRecoveryAfterCrash [gw6] PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node [gw3] PASSED test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml [gw3] PASSED test_config_decryption/test.py::test_successful_decryption_yaml [gw1] PASSED test_catboost_evaluate/test.py::testRecoveryAfterCrash test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh test_analyzer_compatibility/test.py::test_two_new_versions [gw1] PASSED test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments [gw1] PASSED test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments test_codec_encrypted/test.py::test_different_keys test_backup_restore_on_cluster/test_different_versions.py::test_different_versions [gw7] PASSED test_alter_settings_on_cluster/test.py::test_default_database_on_cluster [gw0] PASSED test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] [gw4] PASSED test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] test_backup_restore_s3/test.py::test_backup_to_s3 [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3 test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable [gw1] PASSED test_codec_encrypted/test.py::test_different_keys [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed [gw7] PASSED test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable [gw4] PASSED test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] [gw6] PASSED test_analyzer_compatibility/test.py::test_two_new_versions [gw3] FAILED test_backup_restore_on_cluster/test_different_versions.py::test_different_versions test_backup_log/test.py::test_backup_log [gw1] PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] test_backup_restore_s3/test.py::test_backup_to_s3_multipart test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop [gw7] PASSED test_backup_log/test.py::test_backup_log [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_multipart test_backup_restore_s3/test.py::test_backup_to_s3_named_collection test_config_reloader_interval/test.py::test_reload_config [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_named_collection test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] test_config_xml_main/test.py::test_xml_main_conf [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart test_backup_restore_s3/test.py::test_backup_to_tar test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar test_backup_restore_s3/test.py::test_backup_to_tar_bz2 [gw1] PASSED test_config_reloader_interval/test.py::test_reload_config [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar_bz2 test_backup_restore_s3/test.py::test_backup_to_tar_gz [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar_gz test_backup_restore_s3/test.py::test_backup_to_tar_lzma [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar_lzma test_backup_restore_s3/test.py::test_backup_to_tar_xz [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar_xz test_backup_restore_s3/test.py::test_backup_to_tar_zst [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_tar_zst test_backup_restore_s3/test.py::test_backup_to_zip [gw4] PASSED test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables [gw0] PASSED test_backup_restore_s3/test.py::test_backup_to_zip test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] [gw7] PASSED test_config_xml_main/test.py::test_xml_main_conf [gw0] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] [gw0] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] test_backup_restore_s3/test.py::test_incremental_backup_append_table_def [gw0] PASSED test_backup_restore_s3/test.py::test_incremental_backup_append_table_def test_backup_restore_s3/test.py::test_user_specific_auth [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore [gw4] PASSED test_async_load_databases/test.py::test_multiple_tables [gw0] PASSED test_backup_restore_s3/test.py::test_user_specific_auth [gw2] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup [gw2] FAILED test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup [gw6] PASSED test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop test_config_corresponding_root/test.py::test_work [gw3] PASSED test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop test_config_xml_full/test.py::test_xml_full_conf [gw6] PASSED test_config_corresponding_root/test.py::test_work [gw8] PASSED test_backup_restore_new/test_cancel_backup.py::test_cancel_backup test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup [gw3] PASSED test_config_xml_full/test.py::test_xml_full_conf [gw8] PASSED test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup test_concurrent_queries_restriction_by_query_kind/test.py::test_insert [gw9] PASSED test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] [gw9] PASSED test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] [gw8] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select [gw8] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_backup_restore_keeper_map/test.py::test_on_cluster[0] [gw9] PASSED test_backup_restore_keeper_map/test.py::test_on_cluster[0] test_backup_restore_keeper_map/test.py::test_on_cluster[1] [gw9] PASSED test_backup_restore_keeper_map/test.py::test_on_cluster[1] [gw5] FAILED test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree [gw5] FAILED test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk [gw5] FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated [gw5] FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated =================================== FAILURES =================================== ___________________________ test_different_versions ____________________________ [gw3] linux -- Python 3.10.12 /usr/bin/python3 def test_different_versions(): new_node.query( "CREATE TABLE tbl" " ON CLUSTER 'cluster_ver'" " (x UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}')" " ORDER BY tuple()" ) new_node.query(f"INSERT INTO tbl VALUES (1)") old_node.query(f"INSERT INTO tbl VALUES (2)") backup_name = new_backup_name() initiator = random_node() print(f"Using {get_node_name(initiator)} as initiator for BACKUP") initiator.query(f"BACKUP TABLE tbl ON CLUSTER 'cluster_ver' TO {backup_name}") new_node.query("DROP TABLE tbl ON CLUSTER 'cluster_ver' SYNC") initiator = random_node() print(f"Using {get_node_name(initiator)} as initiator for RESTORE") initiator.query(f"RESTORE TABLE tbl ON CLUSTER 'cluster_ver' FROM {backup_name}") new_node.query("SYSTEM SYNC REPLICA ON CLUSTER 'cluster_ver' tbl") assert new_node.query("SELECT * FROM tbl ORDER BY x") == TSV([1, 2]) assert old_node.query("SELECT * FROM tbl ORDER BY x") == TSV([1, 2]) # Error NO_ELEMENTS_IN_CONFIG is unrelated. > assert ( new_node.query( "SELECT name, last_error_message FROM system.errors WHERE NOT (" "(name == 'NO_ELEMENTS_IN_CONFIG')" ")" ) == "" ) E assert "NETLINK_ERROR\tCan\\'t receive Netlink response: error -2\n" == '' E + NETLINK_ERROR Can\'t receive Netlink response: error -2 test_backup_restore_on_cluster/test_different_versions.py:105: AssertionError ------------------------------ Captured log setup ------------------------------ 2025-04-02 03:37:25 [ 691 ] INFO : Running tests in /ClickHouse/tests/integration/test_backup_restore_on_cluster/test_different_versions.py (cluster.py:2677, start) 2025-04-02 03:37:25 [ 691 ] DEBUG : Cluster start called. is_up=False (cluster.py:2684, start) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker networks for project roottestbackuprestoreonclusterdifferentversions-gw3 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker containers for project roottestbackuprestoreonclusterdifferentversions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker volumes for project roottestbackuprestoreonclusterdifferentversions-gw3 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Cleanup called (cluster.py:876, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker networks for project roottestbackuprestoreonclusterdifferentversions-gw3 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker containers for project roottestbackuprestoreonclusterdifferentversions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Docker volumes for project roottestbackuprestoreonclusterdifferentversions-gw3 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:37:25 [ 691 ] DEBUG : Command:[docker container list --all --filter name='^/roottestbackuprestoreonclusterdifferentversions-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:120, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Unstopped containers: {} (cluster.py:890, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : No running containers for project: roottestbackuprestoreonclusterdifferentversions-gw3 (cluster.py:904, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Trying to prune unused networks... (cluster.py:910, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Trying to prune unused images... (cluster.py:926, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Command:[docker image prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Images pruned (cluster.py:929, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Trying to prune unused volumes... (cluster.py:935, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:120, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Stdout:3 (cluster.py:144, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Command:[docker volume prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:37:25 [ 691 ] DEBUG : Volumes pruned: 3 (cluster.py:940, cleanup) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup directory for instance: new_node (cluster.py:2697, start) 2025-04-02 03:37:25 [ 691 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4530, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Create directory for common tests configuration (cluster.py:4535, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Copy common configuration from helpers (cluster.py:4555, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Generate and write macros file (cluster.py:4607, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster_different_versions.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/configs/config.d (cluster.py:4643, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/database (cluster.py:4660, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/logs (cluster.py:4671, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4752, create_dir) 2025-04-02 03:37:25 [ 691 ] INFO : external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/backups (cluster.py:4799, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup directory for instance: old_node (cluster.py:2697, start) 2025-04-02 03:37:25 [ 691 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4530, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Create directory for common tests configuration (cluster.py:4535, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Copy common configuration from helpers (cluster.py:4555, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Generate and write macros file (cluster.py:4607, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/backups_disk.xml', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/configs/cluster_different_versions.xml'] to /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/configs/config.d (cluster.py:4643, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/database (cluster.py:4660, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/logs (cluster.py:4671, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4752, create_dir) 2025-04-02 03:37:25 [ 691 ] INFO : external_dir_abs_path=/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/backups (cluster.py:4799, create_dir) 2025-04-02 03:37:25 [ 691 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env (cluster.py:95, _create_env_file) 2025-04-02 03:37:25 [ 691 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-02 03:37:25 [ 691 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-02 03:37:25 [ 691 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-02 03:37:25 [ 691 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-02 03:37:25 [ 691 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-02 03:37:25 [ 691 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/docker-compose.yml pull] (cluster.py:120, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: new_node Skipped - Image is already being pulled by zoo1 (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by zoo1 (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: zoo3 Skipped - Image is already being pulled by zoo1 (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: old_node Pulling (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: zoo1 Pulling (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Pulling fs layer (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Waiting (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Downloading [==================================================>] 32B/32B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Downloading [> ] 302.3kB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Downloading [> ] 100.7kB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [> ] 554.3kB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Downloading [=======> ] 3.646kB/25.48kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Downloading [==================================================>] 25.48kB/25.48kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: zoo1 Pulled (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Downloading [> ] 13.78kB/863.5kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Downloading [==================================================>] 863.5kB/863.5kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Downloading [==================================================>] 116B/116B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Downloading [==================================================>] 368B/368B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Downloading [==================================================>] 3.112kB/3.112kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Downloading [================================================> ] 26.7MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [========> ] 25.49MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [> ] 294.9kB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [====================> ] 58.25MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [=====> ] 2.949MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [=============================> ] 84.31MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [==========> ] 5.603MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [================================> ] 94.81MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [=============> ] 7.668MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [=================================> ] 97.98MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [================> ] 9.142MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [===================================> ] 101.1MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [====================> ] 11.21MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [====================================> ] 104.3MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [============================> ] 15.93MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [========================================> ] 22.12MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [=====================================> ] 107.5MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [======================================> ] 110.6MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [==========================================> ] 23.3MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [=======================================> ] 113.8MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [===============================================> ] 26.25MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [========================================> ] 116.9MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [=========================================> ] 120.1MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [=================================================> ] 27.13MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [==========================================> ] 122.8MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Extracting [==================================================>] 27.51MB/27.51MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3823320faa42 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [> ] 98.3kB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [===========================================> ] 125.9MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [===============> ] 2.654MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Downloading [============================================> ] 129.1MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Verifying Checksum (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Download complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [===========================> ] 4.817MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [========================================> ] 7.078MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [==============================================> ] 8.159MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [===============================================> ] 8.356MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [================================================> ] 8.552MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Extracting [==================================================>] 8.803MB/8.803MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 9992e46614b1 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Extracting [==================================================>] 32B/32B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 4f4fb700ef54 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [> ] 557.1kB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [===> ] 11.14MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [======> ] 18.94MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=========> ] 26.18MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [===========> ] 34.54MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [==============> ] 42.34MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=================> ] 49.02MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [==================> ] 54.03MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [====================> ] 59.05MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [======================> ] 65.73MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [========================> ] 71.3MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [==========================> ] 76.87MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=============================> ] 84.12MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [===============================> ] 90.24MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [==================================> ] 98.6MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=====================================> ] 108.1MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [========================================> ] 117MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [===========================================> ] 125.9MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=============================================> ] 132.6MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [================================================> ] 138.7MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [================================================> ] 139.8MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [================================================> ] 140.4MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [================================================> ] 140.9MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [=================================================> ] 142.6MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Extracting [==================================================>] 144.2MB/144.2MB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: cfd7763775f9 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Extracting [==================================================>] 25.48kB/25.48kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Extracting [==================================================>] 25.48kB/25.48kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: a517e5e9f3cd Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Extracting [=> ] 32.77kB/863.5kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Extracting [==================================================>] 863.5kB/863.5kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Extracting [==================================================>] 863.5kB/863.5kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3ce37c9627a5 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Extracting [==================================================>] 116B/116B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Extracting [==================================================>] 116B/116B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 3f8a7f45c802 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Extracting [==================================================>] 368B/368B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Extracting [==================================================>] 368B/368B (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: fd4d333a83bf Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Extracting [==================================================>] 3.112kB/3.112kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Extracting [==================================================>] 3.112kB/3.112kB (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: 12e84a19c313 Pull complete (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Stderr: old_node Pulled (cluster.py:146, run_and_check) 2025-04-02 03:37:37 [ 691 ] DEBUG : Setup ZooKeeper (cluster.py:2738, start) 2025-04-02 03:37:37 [ 691 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper1/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper2/coordination', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/log', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/config', '/ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/keeper3/coordination'] (cluster.py:2739, start) 2025-04-02 03:37:37 [ 691 ] DEBUG : Command:[docker compose --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:120, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr:time="2025-04-02T03:37:37Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Network roottestbackuprestoreonclusterdifferentversions-gw3_default Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Network roottestbackuprestoreonclusterdifferentversions-gw3_default Created (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr:time="2025-04-02T03:37:39Z" level=debug msg="otel error" error="" (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Stderr:time="2025-04-02T03:37:39Z" level=debug msg="otel error" error="" (cluster.py:146, run_and_check) 2025-04-02 03:37:39 [ 691 ] DEBUG : Wait ZooKeeper to start (cluster.py:2403, wait_zookeeper_to_start) 2025-04-02 03:37:39 [ 691 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2004, get_instance_ip) 2025-04-02 03:37:39 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:39 [ 691 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.8.4, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:37:39 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:39 [ 691 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:37:39 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:39 [ 691 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:37:39 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:39 [ 691 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:37:39 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:39 [ 691 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:37:40 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:40 [ 691 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] INFO : Connecting to 172.16.8.4(172.16.8.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:37:42 [ 691 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2004, get_instance_ip) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.8.2, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:37:42 [ 691 ] INFO : Connecting to 172.16.8.2(172.16.8.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:37:42 [ 691 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2004, get_instance_ip) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.8.3, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:37:42 [ 691 ] INFO : Connecting to 172.16.8.3(172.16.8.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:37:42 [ 691 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:37:42 [ 691 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:37:42 [ 691 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:37:42 [ 691 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:37:42 [ 691 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2419, wait_zookeeper_nodes_to_start) 2025-04-02 03:37:42 [ 691 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/docker-compose.yml up -d --no-recreate') (cluster.py:3066, start) 2025-04-02 03:37:42 [ 691 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/docker-compose.yml up -d --no-recreate] (cluster.py:120, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:37:42 [ 691 ] DEBUG : ClickHouse instance created (cluster.py:3074, start) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_instance_ip instance_name=new_node (cluster.py:2004, get_instance_ip) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : get_instance_ip instance_name=new_node (cluster.py:2014, get_instance_global_ipv6) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : Waiting for ClickHouse start in new_node, ip: 172.16.8.5... (cluster.py:3082, start) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/34f28de559101c5a04e7a79750157f9c9875bfd810d7b57fcaa01e63546a7068/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:42 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/34f28de559101c5a04e7a79750157f9c9875bfd810d7b57fcaa01e63546a7068/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/34f28de559101c5a04e7a79750157f9c9875bfd810d7b57fcaa01e63546a7068/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/34f28de559101c5a04e7a79750157f9c9875bfd810d7b57fcaa01e63546a7068/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/34f28de559101c5a04e7a79750157f9c9875bfd810d7b57fcaa01e63546a7068/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : ClickHouse new_node started (cluster.py:3086, start) 2025-04-02 03:37:43 [ 691 ] DEBUG : get_instance_ip instance_name=old_node (cluster.py:2004, get_instance_ip) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : get_instance_ip instance_name=old_node (cluster.py:2014, get_instance_global_ipv6) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : Waiting for ClickHouse start in old_node, ip: 172.16.8.6... (cluster.py:3082, start) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f833575d27dd4b973607b35c13b07b47f29016208808fc339764e7cd05feaad/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f833575d27dd4b973607b35c13b07b47f29016208808fc339764e7cd05feaad/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:37:43 [ 691 ] DEBUG : ClickHouse old_node started (cluster.py:3086, start) ------------------------------ Captured log call ------------------------------- 2025-04-02 03:37:43 [ 691 ] DEBUG : Executing query CREATE TABLE tbl ON CLUSTER 'cluster_ver' (x UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() on new_node (cluster.py:3564, query) 2025-04-02 03:37:43 [ 691 ] DEBUG : Executing query INSERT INTO tbl VALUES (1) on new_node (cluster.py:3564, query) 2025-04-02 03:37:43 [ 691 ] DEBUG : Executing query INSERT INTO tbl VALUES (2) on old_node (cluster.py:3564, query) 2025-04-02 03:37:43 [ 691 ] DEBUG : Executing query BACKUP TABLE tbl ON CLUSTER 'cluster_ver' TO Disk('backups', '1') on new_node (cluster.py:3564, query) 2025-04-02 03:37:43 [ 691 ] DEBUG : Executing query DROP TABLE tbl ON CLUSTER 'cluster_ver' SYNC on new_node (cluster.py:3564, query) 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query RESTORE TABLE tbl ON CLUSTER 'cluster_ver' FROM Disk('backups', '1') on new_node (cluster.py:3564, query) 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query SYSTEM SYNC REPLICA ON CLUSTER 'cluster_ver' tbl on new_node (cluster.py:3564, query) 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query SELECT * FROM tbl ORDER BY x on new_node (cluster.py:3564, query) 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query SELECT * FROM tbl ORDER BY x on old_node (cluster.py:3564, query) 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query SELECT name, last_error_message FROM system.errors WHERE NOT ((name == 'NO_ELEMENTS_IN_CONFIG')) on new_node (cluster.py:3564, query) ---------------------------- Captured log teardown ----------------------------- 2025-04-02 03:37:44 [ 691 ] DEBUG : Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster_ver' SYNC on new_node (cluster.py:3564, query) 2025-04-02 03:37:45 [ 691 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/docker-compose.yml stop --timeout 20] (cluster.py:120, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 03:37:54 [ 691 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/.env --project-name roottestbackuprestoreonclusterdifferentversions-gw3 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/new_node/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-different_versions-0-gw3/old_node/docker-compose.yml down --volumes] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-old_node-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-new_node-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo1-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo2-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Container roottestbackuprestoreonclusterdifferentversions-gw3-zoo3-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Network roottestbackuprestoreonclusterdifferentversions-gw3_default Removing (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stderr: Network roottestbackuprestoreonclusterdifferentversions-gw3_default Removed (cluster.py:146, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Cleanup called (cluster.py:876, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Docker networks for project roottestbackuprestoreonclusterdifferentversions-gw3 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:37:55 [ 691 ] DEBUG : Docker containers for project roottestbackuprestoreonclusterdifferentversions-gw3 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:37:55 [ 691 ] DEBUG : Docker volumes for project roottestbackuprestoreonclusterdifferentversions-gw3 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:37:55 [ 691 ] DEBUG : Command:[docker container list --all --filter name='^/roottestbackuprestoreonclusterdifferentversions-gw3-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Unstopped containers: {} (cluster.py:890, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : No running containers for project: roottestbackuprestoreonclusterdifferentversions-gw3 (cluster.py:904, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Trying to prune unused networks... (cluster.py:910, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Trying to prune unused images... (cluster.py:926, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Command:[docker image prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Images pruned (cluster.py:929, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Trying to prune unused volumes... (cluster.py:935, cleanup) 2025-04-02 03:37:55 [ 691 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stdout:3 (cluster.py:144, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Command:[docker volume prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:37:55 [ 691 ] DEBUG : Volumes pruned: 3 (cluster.py:940, cleanup) _________________________ test_shutdown_cancels_backup _________________________ [gw2] linux -- Python 3.10.12 /usr/bin/python3 def test_shutdown_cancels_backup(): > with NoTrashChecker() as no_trash_checker: test_backup_restore_on_cluster/test_cancel_backup.py:556: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = type = None, value = None, traceback = None def __exit__(self, type, value, traceback): list_of_znodes = set( node1.query( "SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' " + "AND NOT (name == 'alive_tracker')" ).splitlines() ) new_znodes = list_of_znodes.difference(self.__previous_list_of_znodes) if new_znodes: print(f"Found nodes in ZooKeeper: {new_znodes}") for node in new_znodes: print( f"Nodes in '/clickhouse/backups/{node}':\n" + node1.query( f"SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups/{node}'" ) ) print( f"Nodes in '/clickhouse/backups/{node}/stage':\n" + node1.query( f"SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups/{node}/stage'" ) ) if self.check_zookeeper: assert new_znodes == set() list_of_backups = set( os.listdir(os.path.join(node1.cluster.instances_dir, "backups")) ) new_backups = list_of_backups.difference(self.__previous_list_of_backups) unfinished_backups = set( backup for backup in new_backups if not os.path.exists( os.path.join(node1.cluster.instances_dir, "backups", backup, ".backup") ) ) new_backups = set( backup for backup in new_backups if backup not in unfinished_backups ) if new_backups: print(f"Found new backups: {new_backups}") if unfinished_backups: print(f"Found unfinished backups: {unfinished_backups}") assert new_backups == set(self.expect_backups) assert unfinished_backups.difference(self.allow_unfinished_backups) == set() all_errors = set() start_time = time.strftime( "%Y-%m-%d %H:%M:%S", self.__start_time_for_collecting_errors ) for node in nodes: errors_query_result = node.query( "SELECT name FROM system.errors WHERE last_error_time >= toDateTime('" + start_time + "') " + "AND NOT ((name == 'KEEPER_EXCEPTION') AND (last_error_message LIKE '%Fault injection%')) " + "AND NOT (name == 'NO_ELEMENTS_IN_CONFIG')" ) errors = errors_query_result.splitlines() if errors: print(f"{get_node_name(node)}: Found errors: {errors}") print( node.query( "SELECT name, last_error_message FROM system.errors WHERE last_error_time >= toDateTime('" + start_time + "')" ) ) for error in errors: > assert (error in self.expect_errors) or (error in self.allow_errors) E AssertionError: assert ('NETLINK_ERROR' in ['QUERY_WAS_CANCELLED'] or 'NETLINK_ERROR' in []) E + where ['QUERY_WAS_CANCELLED'] = .expect_errors E + and [] = .allow_errors test_backup_restore_on_cluster/test_cancel_backup.py:394: AssertionError ------------------------------ Captured log call ------------------------------- 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' AND NOT (name == 'alive_tracker') on node1 (cluster.py:3564, query) 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query CREATE TABLE tbl ON CLUSTER 'cluster' (x UInt64) ENGINE=ReplicatedMergeTree('/clickhouse/tables/tbl/', '{replica}') ORDER BY tuple() PARTITION BY x%10 on node1 (cluster.py:3564, query) 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query INSERT INTO tbl SELECT number FROM numbers(10) on node1 (cluster.py:3564, query) 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query BACKUP TABLE tbl ON CLUSTER 'cluster' TO Disk('backups', '84ec44ed0ecd4b20a3b7d226169a310c') SETTINGS id='84ec44ed0ecd4b20a3b7d226169a310c' ASYNC on node2 (cluster.py:3564, query) 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query SELECT status FROM system.backups WHERE id='84ec44ed0ecd4b20a3b7d226169a310c' on node2 (cluster.py:3564, query) 2025-04-02 03:38:35 [ 688 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%84ec44ed0ecd4b20a3b7d226169a310c%') on node2 (cluster.py:3564, query) 2025-04-02 03:38:36 [ 688 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%84ec44ed0ecd4b20a3b7d226169a310c%') on node1 (cluster.py:3564, query) 2025-04-02 03:38:37 [ 688 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%84ec44ed0ecd4b20a3b7d226169a310c%') on node1 (cluster.py:3564, query) 2025-04-02 03:38:37 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:37 [ 688 ] DEBUG : Command:[docker exec -u root roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps -C clickhouse] (cluster.py:120, run_and_check) 2025-04-02 03:38:37 [ 688 ] DEBUG : Stdout: PID TTY TIME CMD (cluster.py:144, run_and_check) 2025-04-02 03:38:37 [ 688 ] DEBUG : Stdout: 10 ? 00:00:07 clickhouse (cluster.py:144, run_and_check) 2025-04-02 03:38:37 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:37 [ 688 ] DEBUG : Command:[docker exec -u root roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c pkill clickhouse] (cluster.py:120, run_and_check) 2025-04-02 03:38:37 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:37 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:37 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:38 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:38 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:38 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:39 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:39 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:39 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:40 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:40 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:40 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:41 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:41 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:41 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:42 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:42 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:42 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:43 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:43 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:43 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:44 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:44 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:44 [ 688 ] DEBUG : Stdout:10 (cluster.py:144, run_and_check) 2025-04-02 03:38:45 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:45 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:45 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:45 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:45 [ 688 ] DEBUG : No clickhouse process running. Start new one. (cluster.py:3930, start_clickhouse) 2025-04-02 03:38:45 [ 688 ] DEBUG : http://localhost:None "POST /v1.46/containers/roottestbackuprestoreonclustercancelbackup-gw2-node1-1/exec HTTP/1.1" 201 74 (connectionpool.py:547, _make_request) 2025-04-02 03:38:45 [ 688 ] DEBUG : http://localhost:None "POST /v1.46/exec/ec38b4c0f4d6b2194618537138d2348b7c9ab7ad509065873de4f9c3af9ed07d/start HTTP/1.1" 200 0 (connectionpool.py:547, _make_request) 2025-04-02 03:38:45 [ 688 ] DEBUG : http://localhost:None "GET /v1.46/exec/ec38b4c0f4d6b2194618537138d2348b7c9ab7ad509065873de4f9c3af9ed07d/json HTTP/1.1" 200 586 (connectionpool.py:547, _make_request) 2025-04-02 03:38:46 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:46 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:46 [ 688 ] DEBUG : Stdout:866 (cluster.py:144, run_and_check) 2025-04-02 03:38:46 [ 688 ] DEBUG : Clickhouse process running. (cluster.py:3944, start_clickhouse) 2025-04-02 03:38:46 [ 688 ] DEBUG : run container_id:roottestbackuprestoreonclustercancelbackup-gw2-node1-1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2050, exec_in_container) 2025-04-02 03:38:46 [ 688 ] DEBUG : Command:[docker exec roottestbackuprestoreonclustercancelbackup-gw2-node1-1 bash -c ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:46 [ 688 ] DEBUG : Stdout:866 (cluster.py:144, run_and_check) 2025-04-02 03:38:46 [ 688 ] DEBUG : Executing query select 20 on node1 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%84ec44ed0ecd4b20a3b7d226169a310c%') on node1 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT count() FROM system.processes WHERE (query_kind='Backup') AND (query LIKE '%84ec44ed0ecd4b20a3b7d226169a310c%') on node2 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT status FROM system.backups WHERE id='84ec44ed0ecd4b20a3b7d226169a310c' on node2 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT error FROM system.backups WHERE id='84ec44ed0ecd4b20a3b7d226169a310c' on node2 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SYSTEM FLUSH LOGS on node2 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT status FROM system.backup_log WHERE id='84ec44ed0ecd4b20a3b7d226169a310c' ORDER BY status on node2 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT name FROM system.zookeeper WHERE path = '/clickhouse/backups' AND NOT (name == 'alive_tracker') on node1 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT name FROM system.errors WHERE last_error_time >= toDateTime('2025-04-02 03:38:35') AND NOT ((name == 'KEEPER_EXCEPTION') AND (last_error_message LIKE '%Fault injection%')) AND NOT (name == 'NO_ELEMENTS_IN_CONFIG') on node1 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query SELECT name, last_error_message FROM system.errors WHERE last_error_time >= toDateTime('2025-04-02 03:38:35') on node1 (cluster.py:3564, query) ---------------------------- Captured log teardown ----------------------------- 2025-04-02 03:38:47 [ 688 ] DEBUG : Executing query DROP TABLE IF EXISTS tbl ON CLUSTER 'cluster' SYNC on node1 (cluster.py:3564, query) 2025-04-02 03:38:47 [ 688 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/.env --project-name roottestbackuprestoreonclustercancelbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node2/docker-compose.yml stop --timeout 20] (cluster.py:120, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/.env --project-name roottestbackuprestoreonclustercancelbackup-gw2 --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_backup_restore_on_cluster/_instances-cancel_backup-0-gw2/node2/docker-compose.yml down --volumes] (cluster.py:120, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node2-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-node1-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo2-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo1-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Container roottestbackuprestoreonclustercancelbackup-gw2-zoo3-1 Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw2_default Removing (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Stderr: Network roottestbackuprestoreonclustercancelbackup-gw2_default Removed (cluster.py:146, run_and_check) 2025-04-02 03:38:49 [ 688 ] DEBUG : Cleanup called (cluster.py:876, cleanup) 2025-04-02 03:38:49 [ 688 ] DEBUG : Docker networks for project roottestbackuprestoreonclustercancelbackup-gw2 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:38:49 [ 688 ] DEBUG : Docker containers for project roottestbackuprestoreonclustercancelbackup-gw2 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:38:49 [ 688 ] DEBUG : Docker volumes for project roottestbackuprestoreonclustercancelbackup-gw2 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:38:49 [ 688 ] DEBUG : Command:[docker container list --all --filter name='^/roottestbackuprestoreonclustercancelbackup-gw2-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:120, run_and_check) 2025-04-02 03:38:50 [ 688 ] DEBUG : Unstopped containers: {} (cluster.py:890, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : No running containers for project: roottestbackuprestoreonclustercancelbackup-gw2 (cluster.py:904, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : Trying to prune unused networks... (cluster.py:910, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : Trying to prune unused images... (cluster.py:926, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : Command:[docker image prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:38:50 [ 688 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:38:50 [ 688 ] DEBUG : Images pruned (cluster.py:929, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : Trying to prune unused volumes... (cluster.py:935, cleanup) 2025-04-02 03:38:50 [ 688 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:120, run_and_check) 2025-04-02 03:38:50 [ 688 ] DEBUG : Stdout:1 (cluster.py:144, run_and_check) 2025-04-02 03:38:50 [ 688 ] DEBUG : Volumes pruned: 1 (cluster.py:940, cleanup) _____________________________ test_all_replicated ______________________________ [gw5] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:126: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3600: in query_with_retry result = self.query( helpers/cluster.py:3565: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log setup ------------------------------ 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Stdout:1 (cluster.py:144, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : No running containers (conftest.py:95, cleanup_environment) 2025-04-02 03:36:38 [ 697 ] DEBUG : Pruning Docker networks (conftest.py:97, cleanup_environment) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker network prune --force] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:144, run_and_check) 2025-04-02 03:36:38 [ 697 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2677, start) 2025-04-02 03:36:38 [ 697 ] DEBUG : Cluster start called. is_up=False (cluster.py:2684, start) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw5 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw5 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Cleanup called (cluster.py:876, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw5 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw5 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Unstopped containers: {} (cluster.py:890, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw5 (cluster.py:904, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Trying to prune unused networks... (cluster.py:910, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Trying to prune unused images... (cluster.py:926, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker image prune -f] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Images pruned (cluster.py:929, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Trying to prune unused volumes... (cluster.py:935, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:120, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Stdout:1 (cluster.py:144, run_and_check) 2025-04-02 03:36:38 [ 697 ] DEBUG : Volumes pruned: 1 (cluster.py:940, cleanup) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2697, start) 2025-04-02 03:36:38 [ 697 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4530, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Create directory for common tests configuration (cluster.py:4535, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Copy common configuration from helpers (cluster.py:4555, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Generate and write macros file (cluster.py:4607, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/configs/config.d (cluster.py:4643, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/database (cluster.py:4660, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/logs (cluster.py:4671, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4752, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2697, start) 2025-04-02 03:36:38 [ 697 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4530, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Create directory for common tests configuration (cluster.py:4535, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Copy common configuration from helpers (cluster.py:4555, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Generate and write macros file (cluster.py:4607, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/configs/config.d (cluster.py:4643, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/database (cluster.py:4660, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/logs (cluster.py:4671, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4752, create_dir) 2025-04-02 03:36:38 [ 697 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:8b2301119731', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env (cluster.py:95, _create_env_file) 2025-04-02 03:36:38 [ 697 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-02 03:36:38 [ 697 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-02 03:36:38 [ 697 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-02 03:36:38 [ 697 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-02 03:36:38 [ 697 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-02 03:36:38 [ 697 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --project-name roottestattachpartitionusingcopy-gw5 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/docker-compose.yml pull] (cluster.py:120, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by zoo3 (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by zoo3 (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by zoo3 (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: replica1 Skipped - Image is already being pulled by zoo3 (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: zoo3 Pulling (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Stderr: zoo3 Pulled (cluster.py:146, run_and_check) 2025-04-02 03:36:49 [ 697 ] DEBUG : Setup ZooKeeper (cluster.py:2738, start) 2025-04-02 03:36:49 [ 697 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/keeper3/coordination'] (cluster.py:2739, start) 2025-04-02 03:36:49 [ 697 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy-gw5 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:120, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr:time="2025-04-02T03:36:49Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw5_default Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw5_default Created (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr:time="2025-04-02T03:36:51Z" level=debug msg="otel error" error="" (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Stderr:time="2025-04-02T03:36:51Z" level=debug msg="otel error" error="" (cluster.py:146, run_and_check) 2025-04-02 03:36:51 [ 697 ] DEBUG : Wait ZooKeeper to start (cluster.py:2403, wait_zookeeper_to_start) 2025-04-02 03:36:51 [ 697 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2004, get_instance_ip) 2025-04-02 03:36:51 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:51 [ 697 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.7.2, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:36:51 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:51 [ 697 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:36:51 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:51 [ 697 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:36:52 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:52 [ 697 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:36:52 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:52 [ 697 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:36:53 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:53 [ 697 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-02 03:36:54 [ 697 ] INFO : Connecting to 172.16.7.2(172.16.7.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:36:54 [ 697 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:36:54 [ 697 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:36:54 [ 697 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:36:54 [ 697 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:36:54 [ 697 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2004, get_instance_ip) 2025-04-02 03:36:54 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:54 [ 697 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.7.3, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:36:54 [ 697 ] INFO : Connecting to 172.16.7.3(172.16.7.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:36:54 [ 697 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:36:54 [ 697 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:36:54 [ 697 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:36:54 [ 697 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:36:55 [ 697 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:36:55 [ 697 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:36:55 [ 697 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2004, get_instance_ip) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.7.4, port:2181, use_ssl:False (cluster.py:3223, get_kazoo_client) 2025-04-02 03:36:55 [ 697 ] INFO : Connecting to 172.16.7.4(172.16.7.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-02 03:36:55 [ 697 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=30000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-02 03:36:55 [ 697 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-02 03:36:55 [ 697 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-02 03:36:55 [ 697 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-02 03:36:55 [ 697 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-02 03:36:55 [ 697 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-02 03:36:55 [ 697 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-02 03:36:55 [ 697 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-02 03:36:55 [ 697 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-02 03:36:55 [ 697 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-02 03:36:55 [ 697 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2419, wait_zookeeper_nodes_to_start) 2025-04-02 03:36:55 [ 697 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --project-name roottestattachpartitionusingcopy-gw5 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3066, start) 2025-04-02 03:36:55 [ 697 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --project-name roottestattachpartitionusingcopy-gw5 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:120, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Running (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Creating (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Created (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Starting (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Started (cluster.py:146, run_and_check) 2025-04-02 03:36:55 [ 697 ] DEBUG : ClickHouse instance created (cluster.py:3074, start) 2025-04-02 03:36:55 [ 697 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2004, get_instance_ip) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2014, get_instance_global_ipv6) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.7.6... (cluster.py:3082, start) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/009b7d17467ab0db5c9ab2e7511f5e39b631ffd0d9abefd8003039bafd317d8f/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/009b7d17467ab0db5c9ab2e7511f5e39b631ffd0d9abefd8003039bafd317d8f/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/009b7d17467ab0db5c9ab2e7511f5e39b631ffd0d9abefd8003039bafd317d8f/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:55 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/009b7d17467ab0db5c9ab2e7511f5e39b631ffd0d9abefd8003039bafd317d8f/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/009b7d17467ab0db5c9ab2e7511f5e39b631ffd0d9abefd8003039bafd317d8f/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : ClickHouse replica1 started (cluster.py:3086, start) 2025-04-02 03:36:56 [ 697 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2004, get_instance_ip) 2025-04-02 03:36:56 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2014, get_instance_global_ipv6) 2025-04-02 03:36:56 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.7.5... (cluster.py:3082, start) 2025-04-02 03:36:56 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw5-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : http://localhost:None "GET /v1.46/containers/eafd5c5102bb0fb2723305df42fef68d16f799aefac8d06bb64b0c8528968f9e/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-02 03:36:56 [ 697 ] DEBUG : ClickHouse replica2 started (cluster.py:3086, start) ------------------------------ Captured log call ------------------------------- 2025-04-02 03:36:56 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3564, query) 2025-04-02 03:36:56 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3564, query) 2025-04-02 03:36:56 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3564, query) 2025-04-02 03:36:56 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3564, query) 2025-04-02 03:36:56 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:37:50 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:38:44 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:39:39 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:40:34 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:41:29 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:42:26 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:43:24 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:44:18 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:45:13 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:46:08 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:47:03 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:47:58 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:48:53 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:49:48 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:50:43 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:51:37 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) _____________________________ test_both_mergetree ______________________________ [gw5] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:104: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3600: in query_with_retry result = self.query( helpers/cluster.py:3565: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-02 03:51:38 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3564, query) 2025-04-02 03:52:32 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3564, query) 2025-04-02 03:52:32 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3564, query) 2025-04-02 03:52:32 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3564, query) 2025-04-02 03:52:32 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:53:27 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:54:22 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:55:19 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:56:16 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:57:11 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:58:06 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:59:01 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 03:59:56 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:00:51 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:01:48 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:02:45 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:03:40 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:04:35 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:05:32 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:06:29 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) _______________________ test_not_work_on_different_disk ________________________ [gw5] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:197: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3600: in query_with_retry result = self.query( helpers/cluster.py:3565: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-02 04:06:38 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3564, query) 2025-04-02 04:07:24 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3564, query) 2025-04-02 04:07:24 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3564, query) 2025-04-02 04:07:24 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3564, query) 2025-04-02 04:07:24 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:08:17 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:09:13 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:10:08 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:11:04 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:12:02 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:12:57 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:13:52 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:14:49 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:15:46 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:16:41 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:17:36 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:18:31 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:19:28 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:20:25 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:21:20 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) _______________________ test_only_destination_replicated _______________________ [gw5] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3600: in query_with_retry result = self.query( helpers/cluster.py:3565: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ------------------------------ Captured log call ------------------------------- 2025-04-02 04:21:38 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3564, query) 2025-04-02 04:22:15 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3564, query) 2025-04-02 04:22:15 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3564, query) 2025-04-02 04:22:15 [ 697 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3564, query) 2025-04-02 04:22:15 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:23:10 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:24:05 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:25:00 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:25:54 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:26:51 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:27:49 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:28:44 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:29:39 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:30:34 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:31:29 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:32:23 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:33:18 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:34:13 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:35:08 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) 2025-04-02 04:36:05 [ 697 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3564, query) ---------------------------- Captured log teardown ----------------------------- 2025-04-02 04:36:39 [ 697 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --project-name roottestattachpartitionusingcopy-gw5 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/docker-compose.yml stop --timeout 20] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/.env --project-name roottestattachpartitionusingcopy-gw5 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw5/replica2/docker-compose.yml down --volumes] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica1-1 Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-replica2-1 Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Stopping (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Stopped (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo3-1 Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo2-1 Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw5-zoo1-1 Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw5_default Removing (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw5_default Removed (cluster.py:146, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Cleanup called (cluster.py:876, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw5 are NETWORK ID NAME DRIVER SCOPE (cluster.py:855, print_all_docker_pieces) 2025-04-02 04:36:53 [ 697 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw5 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:863, print_all_docker_pieces) 2025-04-02 04:36:53 [ 697 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw5 are DRIVER VOLUME NAME (cluster.py:871, print_all_docker_pieces) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw5-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Unstopped containers: {} (cluster.py:890, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw5 (cluster.py:904, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Trying to prune unused networks... (cluster.py:910, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Trying to prune unused images... (cluster.py:926, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[docker image prune -f] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:144, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Images pruned (cluster.py:929, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Trying to prune unused volumes... (cluster.py:935, cleanup) 2025-04-02 04:36:53 [ 697 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:120, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Stdout:1 (cluster.py:144, run_and_check) 2025-04-02 04:36:53 [ 697 ] DEBUG : Volumes pruned: 1 (cluster.py:940, cleanup) ============================== slowest durations =============================== 900.00s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 900.00s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 900.00s call test_attach_partition_using_copy/test.py::test_both_mergetree 882.30s call test_attach_partition_using_copy/test.py::test_all_replicated 168.96s setup test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] 126.81s call test_backup_restore_new/test_cancel_backup.py::test_cancel_backup 44.36s call test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop 42.80s call test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop 37.17s call test_async_load_databases/test.py::test_async_load_system_database 32.92s call test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore 30.23s teardown test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] 29.06s call test_async_load_databases/test.py::test_dict_get_data 26.48s setup test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] 23.07s teardown test_backup_restore_s3/test.py::test_user_specific_auth 21.06s setup test_analyzer_compatibility/test.py::test_two_new_versions 20.64s call test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup 20.46s setup test_access_control_on_cluster/test.py::test_access_control_on_cluster 20.29s call test_backup_restore_keeper_map/test.py::test_on_cluster[1] 19.59s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup 19.47s call test_backup_restore_keeper_map/test.py::test_on_cluster[0] 18.92s call test_concurrent_queries_restriction_by_query_kind/test.py::test_select 18.46s setup test_catboost_evaluate/test.py::testAmazonModelManyRows 18.24s setup test_backup_restore_on_cluster/test_different_versions.py::test_different_versions 18.14s setup test_backup_restore_keeper_map/test.py::test_on_cluster[0] 17.70s setup test_attach_partition_using_copy/test.py::test_all_replicated 16.92s setup test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop 16.89s call test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk 16.47s call test_config_xml_full/test.py::test_xml_full_conf 16.33s setup test_alter_settings_on_cluster/test.py::test_default_database_on_cluster 16.10s setup test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes 15.67s call test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore 14.91s setup test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop 14.78s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 14.59s call test_config_xml_main/test.py::test_xml_main_conf 14.19s call test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup 13.38s call test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup 13.31s call test_async_load_databases/test.py::test_dependent_tables 12.72s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 12.69s setup test_config_corresponding_root/test.py::test_work 12.61s setup test_composable_protocols/test.py::test_connections 12.24s setup test_async_load_databases/test.py::test_async_load_system_database 12.20s setup test_backup_restore_new/test_cancel_backup.py::test_cancel_backup 12.09s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 11.79s setup test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 11.62s setup test_config_decryption/test.py::test_successful_decryption_xml 11.18s call test_catboost_evaluate/test.py::testAmazonModelManyRows 10.93s teardown test_backup_restore_on_cluster/test_different_versions.py::test_different_versions 10.54s call test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] 10.27s setup test_codec_encrypted/test.py::test_different_keys 10.02s call test_async_load_databases/test.py::test_multiple_tables 9.93s call test_catboost_evaluate/test.py::testRecoveryAfterCrash 9.62s call test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] 9.42s call test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 8.51s setup test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable 8.34s call test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] 7.57s teardown test_analyzer_compatibility/test.py::test_two_new_versions 7.46s setup test_config_reloader_interval/test.py::test_reload_config 7.14s call test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup 6.64s call test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] 6.50s call test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup 6.36s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 6.29s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 6.10s teardown test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable 5.98s teardown test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster 5.79s teardown test_alter_settings_on_cluster/test.py::test_default_database_on_cluster 5.60s teardown test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node 5.36s call test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node 5.35s call test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node 5.25s setup test_backup_log/test.py::test_backup_log 5.10s call test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes 5.05s call test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node 4.79s teardown test_config_decryption/test.py::test_successful_decryption_yaml 4.75s call test_backup_log/test.py::test_backup_log 4.73s teardown test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 4.65s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_select 4.51s call test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash 4.38s teardown test_codec_encrypted/test.py::test_different_keys 4.03s call test_backup_restore_s3/test.py::test_user_specific_auth 2.87s teardown test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop 2.55s call test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart 2.46s teardown test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop 2.28s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup 2.20s call test_backup_restore_s3/test.py::test_backup_to_s3_multipart 2.14s teardown test_config_reloader_interval/test.py::test_reload_config 2.08s teardown test_backup_restore_keeper_map/test.py::test_on_cluster[1] 2.03s call test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] 1.87s call test_backup_restore_s3/test.py::test_incremental_backup_append_table_def 1.82s call test_analyzer_compatibility/test.py::test_two_new_versions 1.79s teardown test_composable_protocols/test.py::test_proxy_1 1.62s teardown test_async_load_databases/test.py::test_multiple_tables 1.59s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 1.52s call test_backup_restore_s3/test.py::test_backup_to_tar_xz 1.49s call test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] 1.41s call test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] 1.41s call test_backup_restore_on_cluster/test_different_versions.py::test_different_versions 1.34s call test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] 1.33s call test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] 1.31s call test_backup_restore_s3/test.py::test_backup_to_tar 1.31s call test_access_control_on_cluster/test.py::test_access_control_on_cluster 1.30s teardown test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 1.26s call test_backup_restore_s3/test.py::test_backup_to_tar_zst 1.24s call test_backup_restore_s3/test.py::test_backup_to_s3 1.22s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup 1.17s call test_alter_settings_on_cluster/test.py::test_default_database_on_cluster 1.16s call test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] 1.14s teardown test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments 1.14s call test_codec_encrypted/test.py::test_different_keys 1.13s call test_config_reloader_interval/test.py::test_reload_config 1.12s call test_backup_restore_s3/test.py::test_backup_to_tar_gz 1.11s call test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] 1.11s call test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] 1.06s call test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] 1.06s call test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] 1.06s call test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] 1.06s teardown test_backup_log/test.py::test_backup_log 1.01s call test_backup_restore_s3/test.py::test_backup_to_zip 1.01s call test_backup_restore_s3/test.py::test_backup_to_tar_bz2 1.01s call test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] 0.99s call test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster 0.98s call test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 0.96s teardown test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup 0.91s call test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] 0.91s call test_backup_restore_s3/test.py::test_backup_to_tar_lzma 0.86s call test_backup_restore_s3/test.py::test_backup_to_s3_named_collection 0.81s call test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] 0.81s call test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] 0.78s call test_composable_protocols/test.py::test_proxy_1 0.72s call test_composable_protocols/test.py::test_connections 0.71s call test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] 0.59s call test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh 0.57s call test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.56s call test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed 0.52s call test_catboost_evaluate/test.py::testModelPathIsNotAConstString 0.49s call test_catboost_evaluate/test.py::testModelUpdate 0.48s call test_access_control_on_cluster/test.py::test_grant_all_on_cluster 0.46s call test_catboost_evaluate/test.py::testNonConstantFeatures 0.45s call test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable 0.33s call test_config_decryption/test.py::test_successful_decryption_yaml 0.26s call test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments 0.26s call test_catboost_evaluate/test.py::testInvalidModelPath 0.26s call test_catboost_evaluate/test.py::testOnNullableFeatures 0.25s call test_catboost_evaluate/test.py::testInvalidLibraryPath 0.25s call test_catboost_evaluate/test.py::testConstantFeatures 0.25s call test_catboost_evaluate/test.py::testAmazonModelSingleRow 0.22s teardown test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node 0.20s call test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString 0.20s call test_catboost_evaluate/test.py::testOnLowCardinalityFeatures 0.19s call test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric 0.17s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore 0.17s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore 0.17s teardown test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node 0.13s call test_config_decryption/test.py::test_successful_decryption_xml 0.12s teardown test_backup_restore_new/test_cancel_backup.py::test_cancel_backup 0.12s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash 0.12s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup 0.12s teardown test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup 0.12s teardown test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes 0.04s call test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_backup_restore_s3/test.py::test_backup_to_zip 0.00s setup test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] 0.00s teardown test_config_xml_full/test.py::test_xml_full_conf 0.00s setup test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] 0.00s setup test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup 0.00s teardown test_config_xml_main/test.py::test_xml_main_conf 0.00s teardown test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar_gz 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar_bz2 0.00s setup test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] 0.00s setup test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] 0.00s setup test_config_xml_full/test.py::test_xml_full_conf 0.00s setup test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup 0.00s teardown test_async_load_databases/test.py::test_async_load_system_database 0.00s teardown test_backup_restore_keeper_map/test.py::test_on_cluster[0] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] 0.00s teardown test_config_corresponding_root/test.py::test_work 0.00s setup test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node 0.00s teardown test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] 0.00s setup test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] 0.00s setup test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar 0.00s setup test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] 0.00s setup test_backup_restore_s3/test.py::test_user_specific_auth 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_named_collection 0.00s teardown test_access_control_on_cluster/test.py::test_access_control_on_cluster 0.00s setup test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s setup test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash 0.00s setup test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] 0.00s setup test_composable_protocols/test.py::test_proxy_1 0.00s setup test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] 0.00s setup test_async_load_databases/test.py::test_dependent_tables 0.00s setup test_access_control_on_cluster/test.py::test_grant_all_on_cluster 0.00s setup test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] 0.00s setup test_catboost_evaluate/test.py::testRecoveryAfterCrash 0.00s setup test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_catboost_evaluate/test.py::testModelPathIsNotAConstString 0.00s teardown test_composable_protocols/test.py::test_http_proxy_1 0.00s setup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup 0.00s teardown test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] 0.00s teardown test_async_load_databases/test.py::test_dependent_tables 0.00s setup test_catboost_evaluate/test.py::testNonConstantFeatures 0.00s setup test_backup_restore_keeper_map/test.py::test_on_cluster[1] 0.00s setup test_catboost_evaluate/test.py::testOnNullableFeatures 0.00s call test_config_corresponding_root/test.py::test_work 0.00s setup test_async_load_databases/test.py::test_dict_get_data 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3 0.00s teardown test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk 0.00s teardown test_backup_restore_s3/test.py::test_incremental_backup_append_table_def 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart 0.00s setup test_catboost_evaluate/test.py::testModelUpdate 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3 0.00s setup test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node 0.00s teardown test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar_bz2 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk 0.00s teardown test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] 0.00s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s setup test_async_load_databases/test.py::test_multiple_tables 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings 0.00s setup test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString 0.00s teardown test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_multipart 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] 0.00s setup test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s teardown test_catboost_evaluate/test.py::testOnLowCardinalityFeatures 0.00s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_select 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] 0.00s teardown test_async_load_databases/test.py::test_dict_get_data 0.00s setup test_backup_restore_s3/test.py::test_incremental_backup_append_table_def 0.00s setup test_catboost_evaluate/test.py::testAmazonModelSingleRow 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar_zst 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings 0.00s setup test_catboost_evaluate/test.py::testConstantFeatures 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] 0.00s setup test_catboost_evaluate/test.py::testInvalidLibraryPath 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar_zst 0.00s setup test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh 0.00s teardown test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] 0.00s teardown test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_zip 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] 0.00s teardown test_config_decryption/test.py::test_successful_decryption_xml 0.00s setup test_config_decryption/test.py::test_successful_decryption_yaml 0.00s setup test_catboost_evaluate/test.py::testInvalidModelPath 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar_xz 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar_lzma 0.00s teardown test_catboost_evaluate/test.py::testInvalidModelPath 0.00s teardown test_catboost_evaluate/test.py::testModelPathIsNotAConstString 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s setup test_catboost_evaluate/test.py::testOnLowCardinalityFeatures 0.00s setup test_backup_restore_s3/test.py::test_backup_to_s3_multipart 0.00s teardown test_catboost_evaluate/test.py::testAmazonModelManyRows 0.00s teardown test_composable_protocols/test.py::test_connections 0.00s setup test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile 0.00s setup test_backup_restore_s3/test.py::test_backup_to_tar 0.00s setup test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster 0.00s setup test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric 0.00s teardown test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString 0.00s teardown test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_named_collection 0.00s teardown test_catboost_evaluate/test.py::testAmazonModelSingleRow 0.00s teardown test_allow_feature_tier/test.py::test_allow_feature_tier_in_user 0.00s teardown test_catboost_evaluate/test.py::testNonConstantFeatures 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar_gz 0.00s teardown test_catboost_evaluate/test.py::testModelUpdate 0.00s teardown test_catboost_evaluate/test.py::testOnNullableFeatures 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar_xz 0.00s teardown test_catboost_evaluate/test.py::testInvalidLibraryPath 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart 0.00s teardown test_access_control_on_cluster/test.py::test_grant_all_on_cluster 0.00s teardown test_catboost_evaluate/test.py::testRecoveryAfterCrash 0.00s teardown test_catboost_evaluate/test.py::testConstantFeatures 0.00s teardown test_backup_restore_s3/test.py::test_backup_to_tar_lzma =========================== short test summary info ============================ FAILED test_backup_restore_on_cluster/test_different_versions.py::test_different_versions FAILED test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Failed... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Failed... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated PASSED test_composable_protocols/test.py::test_connections PASSED test_composable_protocols/test.py::test_http_proxy_1 PASSED test_composable_protocols/test.py::test_proxy_1 PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility PASSED test_access_control_on_cluster/test.py::test_access_control_on_cluster PASSED test_access_control_on_cluster/test.py::test_grant_all_on_cluster PASSED test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node PASSED test_allow_feature_tier/test.py::test_allow_feature_tier_in_user PASSED test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3] PASSED test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile PASSED test_catboost_evaluate/test.py::testAmazonModelManyRows PASSED test_catboost_evaluate/test.py::testAmazonModelSingleRow PASSED test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket] PASSED test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString PASSED test_catboost_evaluate/test.py::testConstantFeatures PASSED test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric PASSED test_catboost_evaluate/test.py::testInvalidLibraryPath PASSED test_catboost_evaluate/test.py::testInvalidModelPath PASSED test_catboost_evaluate/test.py::testModelPathIsNotAConstString PASSED test_catboost_evaluate/test.py::testModelUpdate PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node PASSED test_catboost_evaluate/test.py::testNonConstantFeatures PASSED test_catboost_evaluate/test.py::testOnLowCardinalityFeatures PASSED test_catboost_evaluate/test.py::testOnNullableFeatures PASSED test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node PASSED test_config_decryption/test.py::test_successful_decryption_xml PASSED test_config_decryption/test.py::test_successful_decryption_yaml PASSED test_catboost_evaluate/test.py::testRecoveryAfterCrash PASSED test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh PASSED test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments PASSED test_alter_settings_on_cluster/test.py::test_default_database_on_cluster PASSED test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain] PASSED test_async_load_databases/test.py::test_async_load_system_database PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local] PASSED test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain] PASSED test_backup_restore_s3/test.py::test_backup_to_s3 PASSED test_codec_encrypted/test.py::test_different_keys PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore PASSED test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable PASSED test_async_load_databases/test.py::test_dependent_tables PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single] PASSED test_analyzer_compatibility/test.py::test_two_new_versions PASSED test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single] PASSED test_backup_log/test.py::test_backup_log PASSED test_backup_restore_s3/test.py::test_backup_to_s3_multipart PASSED test_backup_restore_s3/test.py::test_backup_to_s3_named_collection PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket] PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable] PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup PASSED test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart PASSED test_backup_restore_s3/test.py::test_backup_to_tar PASSED test_config_reloader_interval/test.py::test_reload_config PASSED test_backup_restore_s3/test.py::test_backup_to_tar_bz2 PASSED test_backup_restore_s3/test.py::test_backup_to_tar_gz PASSED test_backup_restore_s3/test.py::test_backup_to_tar_lzma PASSED test_backup_restore_s3/test.py::test_backup_to_tar_xz PASSED test_backup_restore_s3/test.py::test_backup_to_tar_zst PASSED test_async_load_databases/test.py::test_dict_get_data PASSED test_backup_restore_s3/test.py::test_backup_to_zip PASSED test_config_xml_main/test.py::test_xml_main_conf PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False] PASSED test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True] PASSED test_backup_restore_s3/test.py::test_incremental_backup_append_table_def PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup PASSED test_async_load_databases/test.py::test_multiple_tables PASSED test_backup_restore_s3/test.py::test_user_specific_auth PASSED test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore PASSED test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop PASSED test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop PASSED test_config_corresponding_root/test.py::test_work PASSED test_backup_restore_new/test_cancel_backup.py::test_cancel_backup PASSED test_config_xml_full/test.py::test_xml_full_conf PASSED test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup PASSED test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0] PASSED test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select PASSED test_backup_restore_keeper_map/test.py::test_on_cluster[0] PASSED test_backup_restore_keeper_map/test.py::test_on_cluster[1] ================== 6 failed, 94 passed in 3617.85s (1:00:17) =================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 528, in subprocess.check_call(cmd, shell=True, bufsize=0) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_41er5q --privileged --dns-search='.' --memory=30709030912 --security-opt seccomp=unconfined --cap-add=SYS_PTRACE --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=8b2301119731 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_access_control_on_cluster/test.py::test_access_control_on_cluster test_access_control_on_cluster/test.py::test_grant_all_on_cluster test_access_control_on_cluster/test.py::test_grant_current_database_on_cluster test_allow_feature_tier/test.py::test_allow_feature_tier_in_general_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings test_allow_feature_tier/test.py::test_allow_feature_tier_in_mergetree_settings_with_old_compatibility test_allow_feature_tier/test.py::test_allow_feature_tier_in_user test_allow_feature_tier/test.py::test_it_is_possible_to_enable_experimental_settings_in_default_profile test_alter_settings_on_cluster/test.py::test_default_database_on_cluster test_alter_update_cast_keep_nullable/test.py::test_cast_keep_nullable test_analyzer_compatibility/test.py::test_two_new_versions test_async_load_databases/test.py::test_async_load_system_database test_async_load_databases/test.py::test_dependent_tables test_async_load_databases/test.py::test_dict_get_data test_async_load_databases/test.py::test_multiple_tables test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated 'test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[0]' 'test_azure_blob_storage_plain_rewritable/test.py::test_insert_select[1048576]' test_backup_log/test.py::test_backup_log 'test_backup_restore_keeper_map/test.py::test_on_cluster[0]' 'test_backup_restore_keeper_map/test.py::test_on_cluster[1]' test_backup_restore_new/test_cancel_backup.py::test_cancel_backup test_backup_restore_new/test_cancel_backup.py::test_shutdown_cancel_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_cancel_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_error_leaves_no_trash test_backup_restore_on_cluster/test_cancel_backup.py::test_long_disconnection_stops_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_backup test_backup_restore_on_cluster/test_cancel_backup.py::test_short_disconnection_doesnt_stop_restore test_backup_restore_on_cluster/test_cancel_backup.py::test_shutdown_cancels_backup test_backup_restore_on_cluster/test_different_versions.py::test_different_versions test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_different_nodes test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_backups_on_same_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_different_node test_backup_restore_on_cluster/test_disallow_concurrency.py::test_concurrent_restores_on_same_node 'test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3]' 'test_backup_restore_s3/test.py::test_backup_from_s3_to_s3_disk_native_copy[from_s3_to_s3_other_bucket]' test_backup_restore_s3/test.py::test_backup_restore_system_tables_with_plain_rewritable_disk 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_local]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_local_to_s3_plain]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_local]' 'test_backup_restore_s3/test.py::test_backup_to_disk[from_s3_to_s3_plain]' test_backup_restore_s3/test.py::test_backup_to_s3 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_multipart]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[native_single]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_multipart]' 'test_backup_restore_s3/test.py::test_backup_to_s3_different_credentials[non_native_single]' test_backup_restore_s3/test.py::test_backup_to_s3_multipart test_backup_restore_s3/test.py::test_backup_to_s3_named_collection 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3]' 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_other_bucket]' 'test_backup_restore_s3/test.py::test_backup_to_s3_native_copy[policy_s3_plain_rewritable]' test_backup_restore_s3/test.py::test_backup_to_s3_native_copy_multipart test_backup_restore_s3/test.py::test_backup_to_tar test_backup_restore_s3/test.py::test_backup_to_tar_bz2 test_backup_restore_s3/test.py::test_backup_to_tar_gz test_backup_restore_s3/test.py::test_backup_to_tar_lzma test_backup_restore_s3/test.py::test_backup_to_tar_xz test_backup_restore_s3/test.py::test_backup_to_tar_zst test_backup_restore_s3/test.py::test_backup_to_zip 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[False-True-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-False-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-False]' 'test_backup_restore_s3/test.py::test_backup_with_fs_cache[True-True-True]' test_backup_restore_s3/test.py::test_incremental_backup_append_table_def test_backup_restore_s3/test.py::test_user_specific_auth test_catboost_evaluate/test.py::testAmazonModelManyRows test_catboost_evaluate/test.py::testAmazonModelSingleRow test_catboost_evaluate/test.py::testCategoricalFeatureMustBeNumericOrString test_catboost_evaluate/test.py::testConstantFeatures test_catboost_evaluate/test.py::testFloatFeatureMustBeNumeric test_catboost_evaluate/test.py::testInvalidLibraryPath test_catboost_evaluate/test.py::testInvalidModelPath test_catboost_evaluate/test.py::testModelPathIsNotAConstString test_catboost_evaluate/test.py::testModelUpdate test_catboost_evaluate/test.py::testNonConstantFeatures test_catboost_evaluate/test.py::testOnLowCardinalityFeatures test_catboost_evaluate/test.py::testOnNullableFeatures test_catboost_evaluate/test.py::testRecoveryAfterCrash test_catboost_evaluate/test.py::testSystemModelsAndModelRefresh test_catboost_evaluate/test.py::testWrongNumberOfFeatureArguments test_cluster_discovery/test.py::test_cluster_discovery_startup_and_stop test_cluster_discovery/test_auxiliary_keeper.py::test_cluster_discovery_with_auxiliary_keeper_startup_and_stop test_codec_encrypted/test.py::test_different_keys test_composable_protocols/test.py::test_connections test_composable_protocols/test.py::test_http_proxy_1 test_composable_protocols/test.py::test_proxy_1 test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_corresponding_root/test.py::test_work test_config_decryption/test.py::test_successful_decryption_xml test_config_decryption/test.py::test_successful_decryption_yaml test_config_hide_in_preprocessed/test.py::test_hide_in_preprocessed test_config_reloader_interval/test.py::test_reload_config test_config_xml_full/test.py::test_xml_full_conf test_config_xml_main/test.py::test_xml_main_conf -vvv -ss" altinityinfra/integration-tests-runner:2165613c5fcd ' returned non-zero exit status 1.